109 research outputs found
Efficient Video Indexing on the Web: A System that Leverages User Interactions with a Video Player
In this paper, we propose a user-based video indexing method, that
automatically generates thumbnails of the most important scenes of an online
video stream, by analyzing users' interactions with a web video player. As a
test bench to verify our idea we have extended the YouTube video player into
the VideoSkip system. In addition, VideoSkip uses a web-database (Google
Application Engine) to keep a record of some important parameters, such as the
timing of basic user actions (play, pause, skip). Moreover, we implemented an
algorithm that selects representative thumbnails. Finally, we populated the
system with data from an experiment with nine users. We found that the
VideoSkip system indexes video content by leveraging implicit users
interactions, such as pause and thirty seconds skip. Our early findings point
toward improvements of the web video player and its thumbnail generation
technique. The VideSkip system could compliment content-based algorithms, in
order to achieve efficient video-indexing in difficult videos, such as lectures
or sports.Comment: 9 pages, 3 figures, UCMedia 2010: 2nd International ICST Conference
on User Centric Medi
Spectral transform simulations of finite amplitude double-diffusive instabilities in two dimensions
Simulations of double-diffusion with a two-dimensional, vertical plane spectral transform model reveal details of finite amplitude behavior in salt finger, interleaving and diffusive instabilities. Within the range of fluid parameters studied (3 \u3c Ï \u3c 10, .1 \u3c r \u3c .5), infinite, fastest-growing fingers are unstable to Holyer\u27s (1984) nonoscillatory instability and are completely disrupted by it. Finite fingers localized on density steps are also disrupted. Initialized density steps are eroded (the gradients reduced). Fluxes and other diagnostic quantities were determined for salt finger fields at statistical stationarity. These fields contain transitory, irregular finger structures. Fluxes decline steeply as Rfp increases. A single point of comparison of buoyancy flux with ocean measurement yielded good agreement. The dependence of flux ratio on the stability parameter is similar to the linear theory prediction for fastest-growing, infinite fingers and does not increase as Rfp approaches 1, in contrast to laboratory measurements. Holyer\u27s (1984) Floquet theory is extended to the case of nonzero, density compensating, horizontal gradients, and, together with the simulation results, encourages the interpretation of the interleaving instability as being sloping salt fingers. A few preliminary simulations of the diffusive regime indicate very complex behavior. A growing oscillatory perturbation can lead to subcritical convective instability. Such motions sharpen initialized density steps. In the presence of a step, unstable motions are supported even when the fluid is linearly stable to both convection and the diffusive mode
Approximately coloring graphs without long induced paths
It is an open problem whether the 3-coloring problem can be solved in
polynomial time in the class of graphs that do not contain an induced path on
vertices, for fixed . We propose an algorithm that, given a 3-colorable
graph without an induced path on vertices, computes a coloring with
many colors. If the input graph is
triangle-free, we only need many
colors. The running time of our algorithm is if the input
graph has vertices and edges
Complexity of Coloring Graphs without Paths and Cycles
Let and denote a path on vertices and a cycle on
vertices, respectively. In this paper we study the -coloring problem for
-free graphs. Maffray and Morel, and Bruce, Hoang and Sawada,
have proved that 3-colorability of -free graphs has a finite forbidden
induced subgraphs characterization, while Hoang, Moore, Recoskie, Sawada, and
Vatshelle have shown that -colorability of -free graphs for
does not. These authors have also shown, aided by a computer search, that
4-colorability of -free graphs does have a finite forbidden induced
subgraph characterization. We prove that for any , the -colorability of
-free graphs has a finite forbidden induced subgraph
characterization. We provide the full lists of forbidden induced subgraphs for
and . As an application, we obtain certifying polynomial time
algorithms for 3-coloring and 4-coloring -free graphs. (Polynomial
time algorithms have been previously obtained by Golovach, Paulusma, and Song,
but those algorithms are not certifying); To complement these results we show
that in most other cases the -coloring problem for -free
graphs is NP-complete. Specifically, for we show that -coloring is
NP-complete for -free graphs when and ; for we show that -coloring is NP-complete for -free graphs
when , ; and additionally, for , we show that
-coloring is also NP-complete for -free graphs if and
. This is the first systematic study of the complexity of the
-coloring problem for -free graphs. We almost completely
classify the complexity for the cases when , and
identify the last three open cases
List coloring in the absence of a linear forest.
The k-Coloring problem is to decide whether a graph can be colored with at most k colors such that no two adjacent vertices receive the same color. The Listk-Coloring problem requires in addition that every vertex u must receive a color from some given set L(u)â{1,âŠ,k}. Let Pn denote the path on n vertices, and G+H and rH the disjoint union of two graphs G and H and r copies of H, respectively. For any two fixed integers k and r, we show that Listk-Coloring can be solved in polynomial time for graphs with no induced rP1+P5, hereby extending the result of HoĂ ng, KamiĆski, Lozin, Sawada and Shu for graphs with no induced P5. Our result is tight; we prove that for any graph H that is a supergraph of P1+P5 with at least 5 edges, already List 5-Coloring is NP-complete for graphs with no induced H
Complexity of token swapping and its variants
AbstractIn the Token Swapping problem we are given a graph with a token placed on each vertex. Each token has exactly one destination vertex, and we try to move all the tokens to their destinations, using the minimum number of swaps, i.e., operations of exchanging the tokens on two adjacent vertices. As the main result of this paper, we show that Token Swapping is W[1]-hard parameterized by the length k of a shortest sequence of swaps. In fact, we prove that, for any computable function f, it cannot be solved in time f(k)no(k/logk) where n is the number of vertices of the input graph, unless the ETH fails. This lower bound almost matches the trivial nO(k)-time algorithm. We also consider two generalizations of the Token Swapping, namely Colored Token Swapping (where the tokens have colors and tokens of the same color are indistinguishable), and Subset Token Swapping (where each token has a set of possible destinations). To complement the hardness result, we prove that even the most general variant, Subset Token Swapping, is FPT in nowhere-dense graph classes. Finally, we consider the complexities of all three problems in very restricted classes of graphs: graphs of bounded treewidth and diameter, stars, cliques, and paths, trying to identify the borderlines between polynomial and NP-hard cases
Lower bounds on multiple sequence alignment using exact 3-way alignment
<p>Abstract</p> <p>Background</p> <p>Multiple sequence alignment is fundamental. Exponential growth in computation time appears to be inevitable when an optimal alignment is required for many sequences. Exact costs of optimum alignments are therefore rarely computed. Consequently much effort has been invested in algorithms for alignment that are heuristic, or explore a restricted class of solutions. These give an upper bound on the alignment cost, but it is equally important to determine the quality of the solution obtained. In the absence of an optimal alignment with which to compare, lower bounds may be calculated to assess the quality of the alignment. As more effort is invested in improving upper bounds (alignment algorithms), it is therefore important to improve lower bounds as well. Although numerous cost metrics can be used to determine the quality of an alignment, many are based on sum-of-pairs (SP) measures and their generalizations.</p> <p>Results</p> <p>Two standard and two new methods are considered for using exact 2-way and 3-way alignments to compute lower bounds on total SP alignment cost; one new method fares well with respect to accuracy, while the other reduces the computation time. The first employs exhaustive computation of exact 3-way alignments, while the second employs an efficient heuristic to compute a much smaller number of exact 3-way alignments. Calculating all 3-way alignments exactly and computing their average improves lower bounds on sum of SP cost in <it>v</it>-way alignments. However judicious selection of a subset of all 3-way alignments can yield a further improvement with minimal additional effort. On the other hand, a simple heuristic to select a random subset of 3-way alignments (a random packing) yields accuracy comparable to averaging all 3-way alignments with substantially less computational effort.</p> <p>Conclusion</p> <p>Calculation of lower bounds on SP cost (and thus the quality of an alignment) can be improved by employing a mixture of 3-way and 2-way alignments.</p
Evolution through segmental duplications and losses : A Super-Reconciliation approach
The classical gene and species tree reconciliation, used to infer the history of gene gain and loss explaining the evolution of gene families, assumes an independent evolution for each family. While this assumption is reasonable for genes that are far apart in the genome, it is not appropriate for genes grouped into syntenic blocks, which are more plausibly the result of a concerted evolution. Here, we introduce the Super-Reconciliation problem which consists in inferring a history of segmental duplication and loss events (involving a set of neighboring genes) leading to a set of present-day syntenies from a single ancestral one. In other words, we extend the traditional Duplication-Loss reconciliation problem of a single gene tree, to a set of trees, accounting for segmental duplications and losses. Existency of a Super-Reconciliation depends on individual gene tree consistency. In addition, ignoring rearrangements implies that existency also depends on gene order consistency. We first show that the problem of reconstructing a most parsimonious Super-Reconciliation, if any, is NP-hard and give an exact exponential-time algorithm to solve it. Alternatively, we show that accounting for rearrangements in the evolutionary model, but still only minimizing segmental duplication and loss events, leads to an exact polynomial-time algorithm. We finally assess time efficiency of the former exponential time algorithm for the Duplication-Loss model on simulated datasets, and give a proof of concept on the opioid receptor genes
- âŠ